Home
AI Tools
AI Models
MCP
AI NEWS
EN
Model Selection
Tags
VRAM optimization
# VRAM optimization
Deepseek R1 Distill Qwen 32B Quantized.w8a8
MIT
INT8 quantized version of DeepSeek-R1-Distill-Qwen-32B, reducing VRAM usage and improving computational efficiency through weight and activation quantization.
Large Language Model
Transformers
D
neuralmagic
2,324
9
Featured Recommended AI Models
Empowering the Future, Your AI Solution Knowledge Base
English
简体中文
繁體中文
にほんご
© 2025
AIbase